Learning to Act Optimally in Partially Observable Multiagent Settings: (Doctoral Consortium)
نویسنده
چکیده
My research is focused on modeling optimal decision making in partially observable multiagent environments. I began with an investigation into the cognitive biases that induce subnormative behavior in humans playing games online in multiagent settings, leveraging well-known computational psychology approaches in modeling humans playing a strategic, sequential game. My subsequent work was in a scalable extension to Monte Carlo exploring starts for POMDPs (MCES-P), where I expanded the theory and algorithm to the multiagent setting. I first introduced a straightforward application with probably approximately correct guarantees (MCESP+PAC), and then introduced a more sample efficient partially model-based framework (MCESIP+PAC) that explicitly modeled the opponent.
منابع مشابه
Adapting Plans through Communication with Unknown Teammates: (Doctoral Consortium)
Coordinating a team of autonomous agents is a challenging problem. Agents must act in such a way that makes progress toward the achievement of a goal while avoiding conflict with their teammates. In information asymmetric domains, it is often necessary to share crucial observations in order to collaborate effectively. In traditional multiagent systems literature, these teams of agents share an ...
متن کاملLearning and Planning in Multiagent POMDPs Using Finite-State Models of Other Agents
My thesis work provides a new framework for planning in multiagent, stochastic, partially observable domains with little knowledge about other agents. The relevance of the contribution lays in the variety of practical applications this approach can help tackling, given the very generic assumptions about the environment and the other agents. In order to cope with this level of generality, Bayesi...
متن کاملReinforcement Learning in Partially Observable Multiagent Settings: Monte Carlo Exploring Policies with PAC Bounds
Perkins’ Monte Carlo exploring starts for partially observable Markov decision processes (MCES-P) integrates Monte Carlo exploring starts into a local search of policy space to offer a template for reinforcement learning that operates under partial observability of the state. In this paper, we generalize the reinforcement learning under partial observability to the self-interested multiagent se...
متن کاملReinforcement Learning for Decentralized Planning Under Uncertainty (Doctoral Consortium)
Decentralized partially-observable Markov decision processes (Dec-POMDPs) are a powerful tool for modeling multi-agent planning and decision-making under uncertainty. Prevalent Dec-POMDP solution techniques require centralized computation given full knowledge of the underlying model. But in real world scenarios, model parameters may not be known a priori, or may be difficult to specify. We prop...
متن کاملMultiagent Expedition with Graphical Models
We investigate a class of multiagent planning problems termed multiagent expedition, where agents move around an open, unknown, partially observable, stochastic, and physical environment, in pursuit of multiple and alternative goals of different utility. Optimal planning in multiagent expedition is highly intractable.We introduce the notion of conditional optimality, decompose the task into a s...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016